Dauphin County
- North America > United States > Pennsylvania > Dauphin County > Harrisburg (0.04)
- North America > United States > Kentucky (0.04)
- Asia > China > Beijing > Beijing (0.04)
Unsupervised decoding of encoded reasoning using language model interpretability
As large language models become increasingly capable, there is growing concern that they may develop reasoning processes that are encoded or hidden from human oversight. To investigate whether current interpretability techniques can penetrate such encoded reasoning, we construct a controlled testbed by fine-tuning a reasoning model (DeepSeek-R1-Distill-Llama-70B) to perform chain-of-thought reasoning in ROT-13 encryption while maintaining intelligible English outputs. We evaluate mechanistic interpretability methods--in particular, logit lens analysis--on their ability to decode the model's hidden reasoning process using only internal activations. We show that logit lens can effectively translate encoded reasoning, with accuracy peaking in intermediate-to-late layers. Finally, we develop a fully unsupervised decoding pipeline that combines logit lens with automated paraphrasing, achieving substantial accuracy in reconstructing complete reasoning transcripts from internal model representations. These findings suggest that current mechanistic interpretability techniques may be more robust to simple forms of encoded reasoning than previously understood. Our work provides an initial framework for evaluating interpretability methods against models that reason in non-human-readable formats, contributing to the broader challenge of maintaining oversight over increasingly capable AI systems.
- North America > United States > Illinois > Sangamon County > Springfield (0.14)
- North America > United States > Illinois > Cook County > Chicago (0.07)
- North America > United States > California > Sacramento County > Sacramento (0.05)
- (22 more...)
- North America > United States > Pennsylvania > Dauphin County (0.05)
- North America > United States > Texas (0.04)
- North America > United States > North Carolina (0.04)
- (3 more...)
- Government (1.00)
- Energy > Power Industry > Utilities > Nuclear (1.00)
Machine learning-based cloud resource allocation algorithms: a comprehensive comparative review
Cloud resource allocation has emerged as a major challenge in modern computing environments, with organizations struggling to manage complex, dynamic workloads while optimizing performance and cost efficiency. Traditional heuristic approaches prove inadequate for handling the multi-objective optimization demands of existing cloud infrastructures. This paper presents a comparative analysis of state-of-the-art artificial intelligence and machine learning algorithms for resource allocation. We systematically evaluate 10 algorithms across four categories: Deep Reinforcement Learning approaches, Neural Network architectures, Traditional Machine Learning enhanced methods, and Multi-Agent systems. Analysis of published results demonstrates significant performance improvements across multiple metrics including makespan reduction, cost optimization, and energy efficiency gains compared to traditional methods. The findings reveal that hybrid architectures combining multiple artificial intelligence and machine learning techniques consistently outperform single-method approaches, with edge computing environments showing the highest deployment readiness. Our analysis provides critical insights for both academic researchers and industry practitioners seeking to implement next-generation cloud resource allocation strategies in increasingly complex and dynamic computing environments.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- North America > Trinidad and Tobago > Trinidad > Arima > Arima (0.04)
- (9 more...)
- Overview (1.00)
- Research Report (0.84)
- Information Technology > Services (1.00)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine (1.00)
- Law (0.93)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Optimization (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
That New Hit Song on Spotify? It Was Made by A.I.
That New Hit Song on Spotify? Aspiring musicians are churning out tracks using generative artificial intelligence. Some are topping the charts. Nick Arter, a thirty-five-year-old in Washington, D.C., never quite managed to become a professional musician the old-fashioned way. He grew up in Harrisburg, Pennsylvania, in a music-loving family.
- North America > United States > Pennsylvania > Dauphin County > Harrisburg (0.24)
- North America > United States > District of Columbia > Washington (0.24)
- North America > United States > New York (0.05)
- (7 more...)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
Explaining Fine Tuned LLMs via Counterfactuals A Knowledge Graph Driven Framework
Wang, Yucheng, Chen, Ziyang, Kabir, Md Faisal
The widespread adoption of Low-Rank Adaptation (LoRA) has enabled large language models (LLMs) to acquire domain-specific knowledge with remarkable efficiency. However, understanding how such a fine-tuning mechanism alters a model's structural reasoning and semantic behavior remains an open challenge. This work introduces a novel framework that explains fine-tuned LLMs via counterfactuals grounded in knowledge graphs. Specifically, we construct BioToolKG, a domain-specific heterogeneous knowledge graph in bioinformatics tools and design a counterfactual-based fine-tuned LLMs explainer (CFFTLLMExplainer) that learns soft masks over graph nodes and edges to generate minimal structural perturbations that induce maximum semantic divergence. Our method jointly optimizes structural sparsity and semantic divergence while enforcing interpretability preserving constraints such as entropy regularization and edge smoothness. We apply this framework to a fine-tuned LLaMA-based LLM and reveal that counterfactual masking exposes the model's structural dependencies and aligns with LoRA-induced parameter shifts. This work provides new insights into the internal mechanisms of fine-tuned LLMs and highlights counterfactual graphs as a potential tool for interpretable AI.
- North America > United States > Pennsylvania > Dauphin County > Harrisburg (0.04)
- Asia > Myanmar > Tanintharyi Region > Dawei (0.04)
- Overview (0.93)
- Research Report > New Finding (0.46)
E-CaTCH: Event-Centric Cross-Modal Attention with Temporal Consistency and Class-Imbalance Handling for Misinformation Detection
Mousavi, Ahmad, Abdollahinejad, Yeganeh, Corizzo, Roberto, Japkowicz, Nathalie, Boukouvalas, Zois
Detecting multimodal misinformation on social media remains challenging due to inconsistencies between modalities, changes in temporal patterns, and substantial class imbalance. Many existing methods treat posts independently and fail to capture the event-level structure that connects them across time and modality. We propose E-CaTCH, an interpretable and scalable framework for robustly detecting misinformation. If needed, E-CaTCH clusters posts into pseudo-events based on textual similarity and temporal proximity, then processes each event independently. Within each event, textual and visual features are extracted using pre-trained BERT and ResNet encoders, refined via intra-modal self-attention, and aligned through bidirectional cross-modal attention. A soft gating mechanism fuses these representations to form contextualized, content-aware embeddings of each post. To model temporal evolution, E-CaTCH segments events into overlapping time windows and uses a trend-aware LSTM, enhanced with semantic shift and momentum signals, to encode narrative progression over time. Classification is performed at the event level, enabling better alignment with real-world misinformation dynamics. To address class imbalance and promote stable learning, the model integrates adaptive class weighting, temporal consistency regularization, and hard-example mining. The total loss is aggregated across all events. Extensive experiments on Fakeddit, IND, and COVID-19 MISINFOGRAPH demonstrate that E-CaTCH consistently outperforms state-of-the-art baselines. Cross-dataset evaluations further demonstrate its robustness, generalizability, and practical applicability across diverse misinformation scenarios.
- Asia > India (0.05)
- North America > United States > District of Columbia > Washington (0.04)
- North America > United States > Pennsylvania > Dauphin County > Harrisburg (0.04)
- Media > News (1.00)
- Health & Medicine > Therapeutic Area > Immunology (0.35)
Enhancing Clinical Text Classification via Fine-Tuned DRAGON Longformer Models
Yang, Mingchuan, Huang, Ziyuan
This study explores the optimization of the DRAGON Longformer base model for clinical text classification, specifically targeting the binary classification of medical case descriptions. A dataset of 500 clinical cases containing structured medical observations was used, with 400 cases for training and 100 for validation. Enhancements to the pre - trained joeranbosma/dragon - longformer - base - mixed - domain model included hyperparameter tuning, domain - specific preprocessing, and architectural adjustments. Key modifications involved increasing sequence length from 512 to 1024 tokens, adjusting learning rates from 1e - 05 to 5e - 06, extending training epochs from 5 to 8, and incorporating specialized medical terminology. The optimized model achieved notable performance gains: accuracy improved from 72.0% to 85.2%, precision from 68.0% to 84.1%, recall from 75.0% to 86.3%, and F1 - score from 71.0% to 85.2%. Statistical analysis confirmed the significance of these improvements (p < .001). The model demonstrated enhanced capability in interpreting medical terminology, anatomical measurements, and clinical observations. These findings contribute to domain - specific language model research and offer practical implications for clinical natural language processing applications. The optimized model ' s strong performance across diverse medical conditions underscores its potential for broad use in healthcare settings. Enhancing Clinical Text Classification via Fine - Tuned DRAGON Longformer Models Introduction Natural language processing (NLP) in healthcare has continued to advance rapidly, revolutionizing the ability to analyze clinical texts and automate the extraction of valuable insights from massive amounts of medical documentation (Khurana, Koli, Khatter, & Singh, 2023). Over the past few years, large language models (LLMs) have emerged as powerful tools for gaining insight from and processing clinical narratives, creating capabilities that have never been seen before in medical text classification, entity recognition, and clinical decision support (Wang et al., 2018). The DRAGON (Deep Representation Analysis for General - domain Ontology Networks) framework was a specialized version of medical text processing out of all these models (Bosma et al., 2025). Beltagy, Peters, and Cohan (2020) state that the DRAGON longformer model, built on top of the Longformer architecture, addresses the quadratic computational complexity issue of traditional transformer models by processing long sequences.
- North America > United States > Massachusetts > Worcester County > Worcester (0.04)
- North America > United States > Pennsylvania > Dauphin County > Harrisburg (0.04)
- Information Technology > Security & Privacy (0.68)
- Health & Medicine > Health Care Technology (0.46)
- Health & Medicine > Diagnostic Medicine (0.46)
- Information Technology > Artificial Intelligence > Natural Language > Text Classification (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (1.00)
- (2 more...)
OpenAI's Sam Altman thanks Sen John Fetterman for 'normalizing hoodies'
Sen. John Fetterman, D-Pa., receives praise for his less-than-formal attire from Sam Altman during a Commerce Committee hearing. Sen. John Fetterman, D-Pa., was one of the final senators to question OpenAI chief Sam Altman during Thursday's Senate Commerce Committee hearing, and the subject of both Three Mile Island and the Democrat's penchant for Carhartt outerwear came up. Fetterman said that as a senator he has been able to meet people with "much more impressive jobs and careers" and that due to Altman's technology, "humans will have a wonderful ability to adapt." He told Altman that some Americans are worried about AI on various levels, and he asked the executive to address it. In response, Altman said he appreciated Fetterman's praise.
- Energy > Power Industry (0.55)
- Media > News (0.37)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.62)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.62)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.62)
Advancing 3D Medical Image Segmentation: Unleashing the Potential of Planarian Neural Networks in Artificial Intelligence
Huang, Ziyuan, Huggins, Kevin, Bellur, Srikar
Author Note Correspondence concerning this article should be addressed to Ziyuan Huang, University of Massachusetts Chan Medical School, 368 Plantation Street, Worcester, MA 01605 . Advancing 3D Medical Image Segmentation: Unleashing the Potential of Planarian Neural Networks in Artificial Intelligence Abstract Our study presents PNN - UNet as a method fo r constructing deep neural networks that replicate the planarian neural network (PNN) structure in the context of 3D medical image data. Planarians typically have a cerebral structure comprising two neural cords, where the cerebrum acts as a coordinator, and the neural cords serve slightly different purposes within the organism's neurological system. Accordingly, PNN - UNet comprises a D eep - UNet and a W ide - UNet as the nerve cords, with a densely connected autoencoder performing the role of the brain. This dist inct architecture offers advantages over both monolithic (UNet) and modular networks (Ensemble - UNet). Our outcomes on a 3D MRI hippocampus dataset, with and without data augmentation, demonstrate that PNN - UNet outperforms the baseline UNet and several othe r UNet variants in image segmentation. Introduction Medical image segmentation using deep learning techniques plays an increasingly crucial role in assisting clinical diagnosis. Every day, hospitals capture exponentially more medical images, making it increasingly difficult to process big data efficiently and effectively. Medical imaging segmentation can be classified into three major categories: 2D, 2.5D, and 3D (Minaee et al., 2021; Zhang et al., 2022) . The 2D method is to segment 3D images slice - by - slice, utilizing 2D slices as training and testing data. For the 2.5D category, segmentation algorithms usually segment 3D images slice - by - slice, adding neighboring slices as additional inputs. Lastly, 3D im ages are cropped and segmented into small cubic images for training and testing. It is important to note that different methods have their advantages and disadvantages in 3D medical image segmentation.
- North America > United States > Massachusetts > Worcester County > Worcester (0.24)
- North America > United States > Pennsylvania > Dauphin County > Harrisburg (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study > Negative Result (0.46)
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)